What circles do you run in Eliezer? I meet a fair number of people who work in AI, (you can say I "work in AI" myself) and so far I can't think of a single person who was sure of a way to build general intelligence. Is this attitude you observe a common one among people who aren't actually doing AI research, but who think about AI?
Oh, I'm not talking about the mainstream AI field. Most of them know better. I mean, say, a random middle or upper-class individual in Silicon Valley, or a random user on an IRC channel.
However, the rule about instantly solving Friendly AI may apply even within the AI field, since it's a more difficult problem.
It's obvious how to build AI. You just add complexity. AIs need complexity. :-)
I've just finished a 3-day training course on TRIZ (http://en.wikipedia.org/wiki/TRIZ) a problem solving technique, one of the recurring themes throughout the course was what to do about all the solutions that come out even before you've figured out what the true problem is you're trying to solve. The advice was to write the solutions down (rather than be diverted by them or try to bat them away), use them to help examine the problem a bit more and then carry on until you have enough information to make useful judgements about all the solutions you've generated; this was very helpful advice. You need to have a sound way of formulating and exploring the problem space, as well as generating solutions, otherwise you'll become too distracted by all the great solutions your brain is generating.
I just want to remark that it is far from obvious on apriori grounds that there is no elegant general AI algorithm that will solve all the other problems quite nicely. We've only learned this by the continued failure to find such an algorithm or anything like it by the AI community and the continued small successes of more specific less elegant approaches.
AI's need Emergence too. Make sure to add some of that to the soup ;^)
X3J13, the ANSI committee that standarised Common Lisp, had many problems to solve. Kent Pitman credits Larry Masinter with imposing the disciple of seperating problem descriptions from proposed solutions and gives insights into what that meant in practise in a post to comp.lang.lisp
The general interest lies in that fact that the X3J13 Issues were all written up and are available on line.
http://www.lispworks.com/documentation/HyperSpec/Front/X3J13Iss.htm
or
http://www.lisp.org/HyperSpec/FrontMatter/X3J13-Issues.html
so if you wish to study how this works there is a resource you can analyse.
I should confess that my interest has been in content not process. I have been reading these issues to learn Common Lisp. Are these pages really a useful resource for scholars wishing to study the separation of problem descriptions from proposed solutions? I don't know.
I think this argument is flawed with respect to the more technology-oriented questions. Most people do not seriously claim to solve AI problems. What most people (like myself) who are slightly educated in the field (I did an undergrad minor in AI, just very simple stuff) will do is they will suggest an approach that they would try if they had to start working on it. Technical questions also usually yield to evidence very quickly whenever it matters, i.e., when someone would start burning money on an implementation. That is not to say some time and resources are not to be saved by using the maxim outlined here.
OTOH, the part about economists is valid, since most people have very strong ideas (usually wrong ones) about what will work, e.g., as a policy. But then again, most people have no way of wasting (other peoples') resources based on these faulty ideas.
No, wait...
The latest of a number of really good posts from you that directly address the concern of this blog. You seem to be really starting to "grok" the terrifying reality of just how biased we are by the very nature of our thought processes, and coming up with good and useful steps to reduce those biases. Nicely done.
This post makes me wonder how much time passed for Eliezer between concluding that a technological singularity was a probable part of the future and deciding that creating an AGI was the best response, and likewise how much time passed between concluding that AGI Friendlyness would be a difficult problem and concluding that working on a theory of AGI Friendlyness was the best response.
Eliezer, I get the impression that your recent blog entries will make me a better rationalist or if not that a better inventor of software, organizational innovations and social arrangements that will help people become better rationalists.
Good stuff, I say.
A surprising number of people I meet seem to know exactly how to build an Artificial General Intelligence, without, say, knowing how to play the guitar or juggle (much easier problems).
Yes, but while those two topics may be interesting to me, other "easy" problems (home and car maintenance, farming) are not so much even though I recognize their importance. I'm not going to learn how to do everything basic before I am going to learn something complicated. Am I?
Is an AI?
And these problems aren't even easy, really. Like the person who knows how to make an AI, one imagines they "know" how to play guitar. There's a competence level and there is a deeper mastery/creation level. I know three chords; I am not .
Unless that was your point.
This advice seems the opposite of, "avoid analysis paralysis." These may be bounding two extremes, neither of which is healthy. Or I may simply be wrong about the relationship.
Playing the guitar has human-aesthetic components so it's a subproblem of Friendly AI, not just AGI. Building an AI that juggles is a valid challenge. As for trying to do it yourself, that quite misses the point. A mathematician may not be able to do high-speed mental arithmetic, but ought to know how to build a calculator.
I remember reading something much like this in I am right and you are wrong by Edward de Bono, who as I recall wrote that we should try to hold on to the "I haven't made my mind up" state much longer than we do, and be prepared to say "I don't know" much more often than we do (I think he even proposed a new word we could use to answer questions with that meant we don't have a reason to think either way yet). This was about 15 years ago so I've probably mis-remembered.
I was a philosophy undergrad at the time, and when I asked my tutors about de Bono, they told me he was a vacuous 'self-help' nitwit I should ignore.
Completely useless methods for building a general intelligence:
Method 1: Put some bacteria on a lifeless planet with liquid water. Wait until one evolves.
Method 2: Find a fertile human of each gender and induce them to mate. Wait nine months.
Luis Enrique, See above about "We Change Our Minds Less Often Than We Think"; my interpretation is that the people are trying to believe that they haven't made up their minds, but they are wrong. That is, they seem to be implementing the (first) advice you mention. Maybe one can come up with more practical advice, but these are very difficult problems to fix, even if you understand the errors. On the other hand, the main part of the post is about a successful intervention.
Constant, regarding "analysis paralysis," keep in mind there are often two separate questions:
How much time should I spend thinking about X?
Given I'm allocating T time to think about X, how should I divide up T among different thought subtasks?
Analysis Paralysis would generally be a problem with (1).
The current blog post applies more to (2). In the Maier example, the participants presumably know they have a sizable chunk of time blocked out, and the experimental group presumably gets better results not by spending more time overall, but because they reserved a good chunk of T to spend learning the problem, without committing right away to a solution.
The notion of delaying proposition of 'solutions' as long as possible seems an excellent technique for group work where stated propositions not only appear prematurely but become entangled with other, perhaps unproductive interpersonal dynamics, and where the energy of the deliberately 'unmade up' group mind can possibly assist the individual to internally change position. The thorny bit for me however, is the individual trying to 'hold that non-thought' - a challenge that is more or less equivalent to stopping, or even slowing the thought process deliberately, which is meditation after all - something we mere mortals haven't found all that easy so far. Indeed, some argue that many of us aren't even aware there is an 'internal dialogue', let alone knowing how to stop it. In other words, it's easy to say don't make up your mind, but not so easy to enact.
It's okay to think up solutions. You just have to write them down and refocus on the problem.
This is how a brainstorming session is supposed to work. The main goal of the facilitator is to keep the group criticism from spinning out of control. Usually, if someone proposes a solution, someone will shout out an objection to it. But we should still be thinking about the problem. Just write down the solution and shush the objection, then return to the problem.
"...human mind is a lot like the human egg, and the human egg has a shut-off device. When one sperm gets in, it shuts down so the next one can't get in. The human mind has a big tendency of the same sort."
Charlie Munger
I agree. I really hate our notion that "you shouldn't bring up a problem unless you have a solution".
It is obvious to anyone that solves problems that we should analyze the problem before letting our minds move on to a solution.
The people advocating that might be confusing analysis with politics. It's annoying when someone criticises your political idea but offers no alternative; it feels (sometimes accurately) that they're disrupting the conversation but offering no input. So in a political debate, a ground rule might be "don't criticise my solution if you don't have a solution of your own".
Rationally, however, that doesn't excuse not assessing the solution. And it's also important to remember that one potential solution is "do nothing" or "carry on doing what we were doing already". So, in most cases, ANY new solution had an alternative solution to which it can be compared.
Are you sure of that citation? I just looked for it in a copy of Dawes's "Rational Choice in an Uncertain World" and again with the full text search in Google books
http://books.google.com/books/about/Rational_choice_in_an_uncertain_world.html?id=rcU1BsfrM2kC
and did not find any mention of Maier's work. Also, though Maier does frequently use the "Changing Work Procedures" problem, I haven't turned up any publication by him that matches this description. (Note that this failure is quite possibly mine; I haven't done an exhaustive search).
-- MarkusQ
I'm thinking perhaps it is this book by Norman R.F. Maier:
Problem Solving Discussions and Conferences, published by McGraw-Hill Education (December 1963).
Does anyone know of more recent journal article on the topic, 'wait before proposing solutions'?
"why, that problem is so incredibly difficult that an actual majority resolve the whole issue within 15 seconds.", "We Change Our Minds Less Often Than We Think" and "Cached Thoughts"...
Right. We don't do a lot of "our" thinking ourselves. We aren't individually sentient, not really. We don't notice it, but the actual thinking is going on in our subcultures. The sad and funny thing is, we don't even try to understand the cognition of our subcultures, when we research cognition.
I think I'm sentient. If you're not sentient, I would surmise that you believe you're lucky enough to be in a competent subculture -- one self-aware enough to bring this realization to you.
Could one devise a series of experiments to show that individuals aren't sentient, but "subcultures" are?
We aren't individually sentient, not really.
We do less thinking that we imagine, but we still think. However, I still argee (to a lesser extent) that (sub)cultures fixed many thoughts of many people.
The sad and funny thing is, we don't even try to understand the cognition of our subcultures, when we research cognition.
I find 2 possible meaning of "we" here, but the sentence is false in both senses:
This is one of the techniques I've always thought sounded really useful, but never had a clear enough picture of to implement for myself. Does anyone have an example (a transcript, or something of the like) of groups and/or individuals successfully discussing a problem for 5 or 10 minutes without proposing any solutions? I have trouble imagining what that would look like.
No transcript. But I do this professionally all the time. Clients frequently come to me with a design in mind for a solution, and it's often important to back them up and get them to tell me what the problem actually is.
Usually, I start with the question "How would you be able to tell that this problem had been solved?" and repeat it two or twenty times in different words until someone actually tries to answer it.
On one occasion I handed a client my pen and asked whether it was a solution to their problem. They looked at me funny and said it wasn't. I asked them how they knew that, and after a while one of them said "well, for one thing, it doesn't do X" and I said "great!", took the pen back, and wrote "has to do X". Then I handed them the pen back and said "OK, suppose I add the ability to do X somehow to this pen. Is it a solution to your problem now?" and after a couple of iterations they got it and started actually telling me what their problem was.
The thing that used to astonish me is how often the proposed solution utterly fails to even address the problem articulated by the same person who proposed the solution. I've come to expect it.
I start with the question "How would you be able to tell that this problem had been solved?" and repeat it two or twenty times in different words until someone actually tries to answer it.
I handed a client my pen and asked whether it was a solution to their problem
Bleakly funny. Thanks for that. I usually retreat (probably with an angry or pained look on my face) when I notice I'm not really being heard. But sometimes it's better to play and explore.
(nods) It's kind of critical in a systems engineering role.
Only vaguely relatedly, one of my favorite lines ever came from my first professional mentor, about a design he was proposing: "It does what you expect, but you have to expect the right things."
Usually, I start with the question "How would you be able to tell that this problem had been solved?" and repeat it two or twenty times in different words until someone actually tries to answer it.
What a true and hilarious depiction of life. I have the exact same problem doing web development. Because the people giving me projects are not IT people they tend to come up with totally dysfunctional solutions. Yet they almost always start by telling me how they want the problem solved. I have to dig to find out what the problem is first but I just ask them "What result do you want?" or "What purpose do you want this to serve?" and say "I can't make it serve the purpose without knowing what the purpose is." That works for me, without me having to ask them 20 times. Then again maybe you're doing projects in radically different contexts all the time, or with completely different people who vary in their ability to see the point in answering that question. I work with a limited number of people and contexts, all of which I understand pretty well, so my problem clarification process is pretty simple.
What purpose do you want this to serve ... I work with a limited number of people and contexts, all of which I understand pretty well, so my problem clarification process is pretty simple..
In my experience as a programmer (who wore all the software-related hats), I found that even when I understood the domain quite well, inquired about the purpose multiple times, and wrote little stories illustrating my interpretation of the users' desires, I could walk away from early usability tests with major changes to the project.
In one particularly memorable instance, I got all the way through making paper prototypes and making pretend e-mails. Then, I convinced my manager to try out the system. The process started in a pre-existing e-mail package and then routed stuff to the proposed custom software. He sat down, opened up the pretend e-mail, and started to save the attached files. At that point, we discovered that there was no need for the custom software and killed the entire project.
I have attempted using this in more casual decision making situations, and the response I get is nearly always something along the lines of "Okay, just let me propose this one solution, we won't get attached to it or anything, just hear me out..."
What do you do in this situation? Let them speak? Ask them to write down their solution, to be discussed later?
Oops... Couldn't resist proposing solutions.
To be perfectly honest, at the time I simply planted my face on the table in front of me a few times. I was at a dinner party with friends of my mother's; I would have sounded extremely condescending otherwise.
Ah yes, status mismatch in a not very rational crowd. Not much you can do there.
There's a comment already asking for more modern articles/citations/research on this topic but in case someone wants to run with this idea in real life, you can find a summary of Norman Maier's research at http://www.iaf-world.org/Libraries/IAF_Journals/Assets_and_Liabilities_in_Group_Problem_Solving.sflb.ashx
The article was written by Norman Maier in 1967 and reprinted in Psychological Review in 1999. For those of you with access to well-funded libraries, the citations are:
And, to be on really solid ground, you'd want the actual source article(s) that the above review refers to. They are:
From which edition of the book does the reference originate? At first glance it does not seem to be included in the second edition and I'm curious to read more about it.
I think the main important lesson is to not get attached to early ideas. Instead of banning early ideas, if anything comes up, you can just write tit down, and set it aside. I find this easier than a full ban, because it's just an easier move to make for my brain.
(I have a similar problem with rationalist taboo. Don't ban words, instead require people to locally define their terms for the duration of the conversation. It solves the same problem, and it isn't a ban on though or speech.)
The other important lesson of the post, is that, in the early discussion, focus on increasing your shared understanding of the problem, rather than generating ideas. I.e. it's ok for ideas to come up (and when they do you save them for later). But generating ideas is not the goal in the beginning.
Hm, thinking about it, I think the mechanism of classical brainstorming (where you up front think of as many ideas as you can) is to exhaust all the trivial, easy to think of, ideas, as fast as you can, and then you're forced to think deeper to come up with new ideas. I guess that's another way to do it. But I think this is method is both ineffective and unreliable, since it only works though a secondary effect.
. . .
It is interesting to comparing the advise in this post with the Game Tree of Aliment or Builder/Breaker Methodology also here. I've seen variants of this exercise popping lots of places in the AI Safety community. Some of them pare probably inspired by each other, but I'm pretty sure (80%) that this method have been invented several times independently.
I think that GTA/BBM works for the same reason the advice in the post works. It also solves the problem of not getting attached, and also as you keep breaking your ideas and explore new territory, you expand your understanding or the problem. I think an active ingrediens in this method is that the people playing this game knows that alignment is hard, and go in expecting their first several ideas to be terrible. You know the exercise is about noticing the flaws in your plans, and learn from your mistakes. Without this attitude, I don't think it would work very well.
From Robyn Dawes’s Rational Choice in an Uncertain World.1 Bolding added.
This is so true it’s not even funny. And it gets worse and worse the tougher the problem becomes. Take artificial intelligence, for example. A surprising number of people I meet seem to know exactly how to build an artificial general intelligence, without, say, knowing how to build an optical character recognizer or a collaborative filtering system (much easier problems). And as for building an AI with a positive impact on the world—a Friendly AI, loosely speaking—why, that problem is so incredibly difficult that an actual majority resolve the whole issue within fifteen seconds.2 Give me a break.
This problem is by no means unique to AI. Physicists encounter plenty of nonphysicists with their own theories of physics, economists get to hear lots of amazing new theories of economics. If you’re an evolutionary biologist, anyone you meet can instantly solve any open problem in your field, usually by postulating group selection. Et cetera.
Maier’s advice echoes the principle of the bottom line, that the effectiveness of our decisions is determined only by whatever evidence and processing we did in first arriving at our decisions—after you write the bottom line, it is too late to write more reasons above. If you make your decision very early on, it will, in fact, be based on very little thought, no matter how many amazing arguments you come up with afterward.
And consider furthermore that we change our minds less often than we think: 24 people assigned an average 66% probability to the future choice thought more probable, but only 1 in 24 actually chose the option thought less probable. Once you can guess what your answer will be, you have probably already decided. If you can guess your answer half a second after hearing the question, then you have half a second in which to be intelligent. It’s not a lot of time.
Traditional Rationality emphasizes falsification—the ability to relinquish an initial opinion when confronted by clear evidence against it. But once an idea gets into your head, it will probably require way too much evidence to get it out again. Worse, we don’t always have the luxury of overwhelming evidence.
I suspect that a more powerful (and more difficult) method is to hold off on thinking of an answer. To suspend, draw out, that tiny moment when we can’t yet guess what our answer will be; thus giving our intelligence a longer time in which to act.
Even half a minute would be an improvement over half a second.
1Robyn M. Dawes, Rational Choice in An Uncertain World, 1st ed., ed. Jerome Kagan (San Diego, CA: Harcourt Brace Jovanovich, 1988), 55–56.
2See Yudkowsky, “Artificial Intelligence as a Positive and Negative Factor in Global Risk.”